Automatic Taxonomy Extraction in Different Languages Using Wikipedia and Minimal Language-Specific Information

نویسندگان

  • Renato Domínguez García
  • Sebastian Schmidt
  • Christoph Rensing
  • Ralf Steinmetz
چکیده

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Language - and domain - independent text mining

Aalto University, P.O. Box 11000, FI-00076 Aalto www.aalto.fi Author Mari-Sanna Paukkeri Name of the doctoral dissertation Languageand domain-independent text mining Publisher School of Science Unit Department of Information and Computer Science Series Aalto University publication series DOCTORAL DISSERTATIONS 137/2012 Field of research Computer and Information Science Manuscript submitted 4 Ma...

متن کامل

Bootstrapping Term Extractors for Multiple Languages

Terminology extraction resources are needed for a wide range of human language technology applications, including knowledge management, information extraction, semantic search, cross-language information retrieval and automatic and assisted translation. We report a low cost method for creating terminology extraction resources for 21 non-English EU languages. Using parallel corpora and a project...

متن کامل

Pazienza University of Roma Tor Vergata , Italy Armando Stellato University of Roma Tor Vergata , Italy Semi - Automatic Ontology Development : Processes and Resources

The collection of the specialized vocabulary of a particular domain (terminology) is an important initial step of creating formalized domain knowledge representations (ontologies). Terminology Extraction (TE) aims at automating this process by collecting the relevant domain vocabulary from existing lexical resources or collections of domain texts. In this chapter, the authors address the extrac...

متن کامل

Information Extraction from Wikipedia Using Pattern Learning

In this paper we present solutions for the crucial task of extracting structured information from massive free-text resources, such as Wikipedia, for the sake of semantic databases serving upcoming Semantic Web technologies. We demonstrate both a verb frame-based approach using deep natural language processing techniques with extraction patterns developed by human knowledge experts and machine ...

متن کامل

Automatic Population of Structured Knowledge Bases via Natural Language Processing

The Web has evolved into a huge mine of knowledge carved in different forms, the predominant one still being the free-text document. This motivates the need for Intelligent Web-reading Agents: hypothetically, they would skim through disparate Web sources corpora and generate meaningful structured assertions to fuel Knowledge Bases (KBs). Ultimately, comprehensive KBs, like Wikidata and DBpedia,...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2012